Learning Compact Features via In-Training Representation Alignment
نویسندگان
چکیده
Deep neural networks (DNNs) for supervised learning can be viewed as a pipeline of the feature extractor (i.e., last hidden layer) and linear classifier output that are trained jointly with stochastic gradient descent (SGD) on loss function (e.g., cross-entropy). In each epoch, true is estimated using mini-batch sampled from training set model parameters then updated gradients. Although latter provides an unbiased estimation former, they subject to substantial variances derived size number mini-batches, leading noisy jumpy updates. To stabilize such undesirable variance in estimating gradients, we propose In-Training Representation Alignment (ITRA) explicitly aligns distributions two different mini-batches matching SGD process. We also provide rigorous analysis desirable effects representation learning: (1) extracting compact representation; (2) reducing over-adaption via adaptively weighting mechanism; (3) accommodating multi-modalities. Finally, conduct large-scale experiments both image text classifications demonstrate its superior performance strong baselines.
منابع مشابه
Learning Via Compact Data Representation
We present an unsupervised learning methodology derived from compact data encoding and demonstrate how to construct models of polysemy, priming, semantic disambiguation and learning using this theoretical basis. The model is capable of simulating human-like performance on artificial grammar learning.
متن کاملImage alignment via kernelized feature learning
Machine learning is an application of artificial intelligence that is able to automatically learn and improve from experience without being explicitly programmed. The primary assumption for most of the machine learning algorithms is that the training set (source domain) and the test set (target domain) follow from the same probability distribution. However, in most of the real-world application...
متن کاملImage Classification via Sparse Representation and Subspace Alignment
Image representation is a crucial problem in image processing where there exist many low-level representations of image, i.e., SIFT, HOG and so on. But there is a missing link across low-level and high-level semantic representations. In fact, traditional machine learning approaches, e.g., non-negative matrix factorization, sparse representation and principle component analysis are employed to d...
متن کاملA Novel Graph-based Compact Representation of Word Alignment
In this paper, we propose a novel compact representation called weighted bipartite hypergraph to exploit the fertility model, which plays a critical role in word alignment. However, estimating the probabilities of rules extracted from hypergraphs is an NP-complete problem, which is computationally infeasible. Therefore, we propose a divide-and-conquer strategy by decomposing a hypergraph into a...
متن کاملLearning Features of Representation in Conceptual Context
When people categorize an object, they often encode a certain number of its properties for later classification. In Schyns and Murphy (1993), we suggested that the way people group objects into categories could induce the learning of new dimensions of ca egorization--i.e., dimensions that did not exist prior to the experience with the categorization system. In this research, we examine whether ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2023
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v37i7.26044